Dynamical stability and chaos in artificial neural network trajectories along training
Abstract: The process of training an artificial neural network involves iteratively adapting its parameters so as to minimize the error of the network's prediction, when confronted with a learning task. This iterative change can be naturally interpreted as a trajectory in network space -- a time series of networks -- and thus the training algorithm (e.g. gradient descent optimization of a suitable loss function) can be interpreted as a dynamical system in graph space. In order to illustrate this interpretation, here we study the dynamical properties of this process by analyzing through this lens the network trajectories of a shallow neural network, and its evolution through learning a simple classification task. We systematically consider different ranges of the learning rate and explore both the dynamical and orbital stability of the resulting network trajectories, finding hints of regular and chaotic behavior depending on the learning rate regime. Our findings are put in contrast to common wisdom on convergence properties of neural networks and dynamical systems theory. This work also contributes to the cross-fertilization of ideas between dynamical systems theory, network theory and machine learning
- Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT press, Cambridge (2016) Marcus [2018] Marcus, G.: Deep Learning: A Critical Appraisal. arXiv (2018) https://doi.org/10.48550/ARXIV.1801.00631 San Miguel [2023] San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Marcus, G.: Deep Learning: A Critical Appraisal. arXiv (2018) https://doi.org/10.48550/ARXIV.1801.00631 San Miguel [2023] San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Marcus, G.: Deep Learning: A Critical Appraisal. arXiv (2018) https://doi.org/10.48550/ARXIV.1801.00631 San Miguel [2023] San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Caligiuri, A., EguÃluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
- Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.